skip to main content


Search for: All records

Creators/Authors contains: "Moore, Sean"

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. Abstract

    Our ability to forecast epidemics far into the future is constrained by the many complexities of disease systems. Realistic longer-term projections may, however, be possible under well-defined scenarios that specify the future state of critical epidemic drivers. Since December 2020, the U.S. COVID-19 Scenario Modeling Hub (SMH) has convened multiple modeling teams to make months ahead projections of SARS-CoV-2 burden, totaling nearly 1.8 million national and state-level projections. Here, we find SMH performance varied widely as a function of both scenario validity and model calibration. We show scenarios remained close to reality for 22 weeks on average before the arrival of unanticipated SARS-CoV-2 variants invalidated key assumptions. An ensemble of participating models that preserved variation between models (using the linear opinion pool method) was consistently more reliable than any single model in periods of valid scenario assumptions, while projection interval coverage was near target levels. SMH projections were used to guide pandemic response, illustrating the value of collaborative hubs for longer-term scenario projections.

     
    more » « less
  2. null (Ed.)
    By March 2020, COVID-19 led to thousands of deaths and disrupted economic activity worldwide. As a result of narrow case definitions and limited capacity for testing, the number of unobserved severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) infections during its initial invasion of the United States remains unknown. We developed an approach for estimating the number of unobserved infections based on data that are commonly available shortly after the emergence of a new infectious disease. The logic of our approach is, in essence, that there are bounds on the amount of exponential growth of new infections that can occur during the first few weeks after imported cases start appearing. Applying that logic to data on imported cases and local deaths in the United States through 12 March, we estimated that 108,689 (95% posterior predictive interval [95% PPI]: 1,023 to 14,182,310) infections occurred in the United States by this date. By comparing the model’s predictions of symptomatic infections with local cases reported over time, we obtained daily estimates of the proportion of symptomatic infections detected by surveillance. This revealed that detection of symptomatic infections decreased throughout February as exponential growth of infections outpaced increases in testing. Between 24 February and 12 March, we estimated an increase in detection of symptomatic infections, which was strongly correlated (median: 0.98; 95% PPI: 0.66 to 0.98) with increases in testing. These results suggest that testing was a major limiting factor in assessing the extent of SARS-CoV-2 transmission during its initial invasion of the United States. 
    more » « less
  3. Ouellette, Francis (Ed.)
    Experimental data about gene functions curated from the primary literature have enormous value for research scientists in understanding biology. Using the Gene Ontology (GO), manual curation by experts has provided an important resource for studying gene function, especially within model organisms. Unprecedented expansion of the scientific literature and validation of the predicted proteins have increased both data value and the challenges of keeping pace. Capturing literature-based functional annotations is limited by the ability of biocurators to handle the massive and rapidly growing scientific literature. Within the community-oriented wiki framework for GO annotation called the Gene Ontology Normal Usage Tracking System (GONUTS), we describe an approach to expand biocuration through crowdsourcing with undergraduates. This multiplies the number of high-quality annotations in international databases, enriches our coverage of the literature on normal gene function, and pushes the field in new directions. From an intercollegiate competition judged by experienced biocurators, Community Assessment of Community Annotation with Ontologies (CACAO), we have contributed nearly 5,000 literature-based annotations. Many of those annotations are to organisms not currently well-represented within GO. Over a 10-year history, our community contributors have spurred changes to the ontology not traditionally covered by professional biocurators. The CACAO principle of relying on community members to participate in and shape the future of biocuration in GO is a powerful and scalable model used to promote the scientific enterprise. It also provides undergraduate students with a unique and enriching introduction to critical reading of primary literature and acquisition of marketable skills. 
    more » « less
  4. Abstract

    The ability to theoretically predict accurate NMR chemical shifts in solids is increasingly important due to the role such shifts play in selecting among proposed model structures. Herein, two theoretical methods are evaluated for their ability to assign15N shifts from guanosine dihydrate to one of the two independent molecules present in the lattice. The NMR data consist of15N shift tensors from 10 resonances. Analysis using periodic boundary or fragment methods consider a benchmark dataset to estimate errors and predict uncertainties of 5.6 and 6.2 ppm, respectively. Despite this high accuracy, only one of the five sites were confidently assigned to a specific molecule of the asymmetric unit. This limitation is not due to negligible differences in experimental data, as most sites exhibit differences of >6.0 ppm between pairs of resonances representing a given position. Instead, the theoretical methods are insufficiently accurate to make assignments at most positions.

     
    more » « less
  5. Abstract Academic researchers, government agencies, industry groups, and individuals have produced forecasts at an unprecedented scale during the COVID-19 pandemic. To leverage these forecasts, the United States Centers for Disease Control and Prevention (CDC) partnered with an academic research lab at the University of Massachusetts Amherst to create the US COVID-19 Forecast Hub. Launched in April 2020, the Forecast Hub is a dataset with point and probabilistic forecasts of incident cases, incident hospitalizations, incident deaths, and cumulative deaths due to COVID-19 at county, state, and national, levels in the United States. Included forecasts represent a variety of modeling approaches, data sources, and assumptions regarding the spread of COVID-19. The goal of this dataset is to establish a standardized and comparable set of short-term forecasts from modeling teams. These data can be used to develop ensemble models, communicate forecasts to the public, create visualizations, compare models, and inform policies regarding COVID-19 mitigation. These open-source data are available via download from GitHub, through an online API, and through R packages. 
    more » « less
  6. Short-term probabilistic forecasts of the trajectory of the COVID-19 pandemic in the United States have served as a visible and important communication channel between the scientific modeling community and both the general public and decision-makers. Forecasting models provide specific, quantitative, and evaluable predictions that inform short-term decisions such as healthcare staffing needs, school closures, and allocation of medical supplies. Starting in April 2020, the US COVID-19 Forecast Hub ( https://covid19forecasthub.org/ ) collected, disseminated, and synthesized tens of millions of specific predictions from more than 90 different academic, industry, and independent research groups. A multimodel ensemble forecast that combined predictions from dozens of groups every week provided the most consistently accurate probabilistic forecasts of incident deaths due to COVID-19 at the state and national level from April 2020 through October 2021. The performance of 27 individual models that submitted complete forecasts of COVID-19 deaths consistently throughout this year showed high variability in forecast skill across time, geospatial units, and forecast horizons. Two-thirds of the models evaluated showed better accuracy than a naïve baseline model. Forecast accuracy degraded as models made predictions further into the future, with probabilistic error at a 20-wk horizon three to five times larger than when predicting at a 1-wk horizon. This project underscores the role that collaboration and active coordination between governmental public-health agencies, academic modeling teams, and industry partners can play in developing modern modeling capabilities to support local, state, and federal response to outbreaks. 
    more » « less